Conference Proceedings

Invertible Concept-based Explanations for CNN Models with Non-negative Concept Activation Vectors

R Zhang, P Madumal, T Miller, KA Ehinger, BIP Rubinstein

35th Aaai Conference on Artificial Intelligence Aaai 2021 | Published : 2021

Abstract

Convolutional neural network (CNN) models for computer vision are powerful but lack explainability in their most basic form. This deficiency remains a key challenge when applying CNNs in important domains. Recent work on explanations through feature importance of approximate linear models has moved from input-level features (pixels or segments) to features from mid-layer feature maps in the form of concept activation vectors (CAVs). CAVs contain concept-level information and could be learned via clustering. In this work, we rethink the ACE algorithm of Ghorbani et al., proposing an alternative invertible concept-based explanation (ICE) framework to overcome its shortcomings. Based on the req..

View full abstract

University of Melbourne Researchers

Grants

Awarded by University of Melbourne


Funding Acknowledgements

This research is supported by Australian Research Council (ARC) Discovery Grant DP190103414: Explanation in Artificial Intelligence: A Human-Centred Approach. The first two authors are supported by the University of Melbourne research scholarship (MRS) scheme. Experiments were undertaken using the LIEF HPC-GPGPU Facility hosted at the University of Melbourne. This Facility was established with the assistance of LIEF Grant LE170100200.